Computing the Kullback-Leibler Divergence between two Generalized Gamma Distributions
نویسنده
چکیده
We derive a closed form solution for the Kullback-Leibler divergence between two generalized gamma distributions. These notes are meant as a reference and provide a guided tour towards a result of practical interest that is rarely explicated in the literature. 1 The Generalized Gamma Distribution The origins of the generalized gamma distribution can be traced back to work of Amoroso in 1925 [1,2]. Here, we are concerned with the three-parameter version that was later introduced by Stacy [3]. Its probability density function is defined for x ∈ [0,∞) and given by f(x | a, d, p) = p ad xd−1 Γ (d/p) exp [ − (x a )p] (1) where Γ (·) is the gamma function, a > 0 determines scale and d > 0 and p > 0 are two shape parameters. We note that, depending on its parametrization, this unimodal density may be skewed to the left or to the right. Moreover, the generalized gamma contains other distributions as special cases. For d = p, it coincides with the Weibull distribution, and, if p = 1, it becomes the gamma distribution. Setting d = p = 1 yields the exponential distribution, and, for a = 2, p = 1, and d = k/2 where k ∈ N, we obtain the χ distribution with k degrees of freedom. As a flexible skewed distribution, the generalized gamma is frequently used for life-time analysis and reliability testing. In addition, it models fading phenomena in wireless communication, has been applied in automatic image retrieval and analysis [4,5,6], was used to evaluate dimensionality reduction techniques [7], and also appears to be connected to diffusion processes in (social) networks [8,9,10]. Accordingly, methods for measuring (dis)similarities between generalized gamma distributions are of practical interest in data science because they facilitate model selection and statistical inference. 2 The Kullback-Leibler Divergence The Kullback-Leibler divergence (KL divergence) provides an asymmetric measure of the similarity of two probability distributions P and Q [11]. For the case where the two distributions are continuous, it is defined as
منابع مشابه
Information Measures via Copula Functions
In applications of differential geometry to problems of parametric inference, the notion of divergence is often used to measure the separation between two parametric densities. Among them, in this paper, we will verify measures such as Kullback-Leibler information, J-divergence, Hellinger distance, -Divergence, … and so on. Properties and results related to distance between probability d...
متن کاملMultivariate Generalized Gamma Distribution for Content Based Image Retrieval
This paper deals with the joint modeling of color textures in the context of Content Based Image Retrieval (CBIR). We propose a generic multivariate model based on the Generalized Gamma distribution to describe the marginal behavior of texture wavelet detail subbands. Then the information of dependence across color components is incorporated in the modeling process using the Gaussian copula. Th...
متن کاملExpected Logarithm of Central Quadratic Form and Its Use in KL-Divergence of Some Distributions
Abstract: In this paper, we develop three different methods for computing the expected logarithm of central quadratic forms: a series method, an integral method and a fast (but inexact) set of methods. The approach used for deriving the integral method is novel and can be used for computing the expected logarithm of other random variables. Furthermore, we derive expressions for the Kullback–Lei...
متن کاملThe predictive Lasso
We propose a shrinkage procedure for simultaneous variable selection and estimation in generalized linear models (GLMs) with an explicit predictive motivation. The procedure estimates the coefficients by minimizing the Kullback-Leibler divergence of a set of predictive distributions to the corresponding predictive distributions for the full model, subject to an constraint on the coefficient vec...
متن کاملJensen divergence based on Fisher's information
The measure of Jensen-Fisher divergence between probability distributions is introduced and its theoretical grounds set up. This quantity, in contrast to the remaining Jensen divergences, is very sensitive to the fluctuations of the probability distributions because it is controlled by the (local) Fisher information, which is a gradient functional of the distribution. So, it is appropriate and ...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید
ثبت ناماگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید
ورودعنوان ژورنال:
- CoRR
دوره abs/1401.6853 شماره
صفحات -
تاریخ انتشار 2014